SMILE: Simulator for Maryland Imitation Learning Environment
نویسندگان
چکیده
As robot imitation learning is beginning to replace conventional hand-coded approaches in programming robot behaviors, much work is focusing on learning from the actions of demonstrators. We hypothesize that in many situations, procedural tasks can be learned more effectively by observing object behaviors while completely ignoring the demonstrator’s motions. To support studying this hypothesis and robot imitation learning in general, we built a software system named smile that is a simulated 3D environment. In this virtual environment, both a simulated robot and a user-controlled demonstrator can manipulate various objects on a tabletop. The demonstrator is not embodied in smile, and therefore a recorded demonstration appears as if the objects move on their own. In addition to recording demonstrations, smile also allows programing the simulated robot via Matlab scripts, as well as creating highly customizable objects for task scenarios via XML. This report describes the features and usages of smile. Acknowledgements: This work was supported by ONR award N000141310597. We thank John Purtilo and Charmi Patel for assistance in the implementation of smile. This report is an update for CS-TR-5039 published in June 2014. ∗Email: [email protected]
منابع مشابه
The Maryland Virtual Demonstrator Environment for Robot Imitation Learning
Robot imitation learning, where a robot autonomously generates actions required to accomplish a task demonstrated by a human, has emerged as a potential replacement for a more conventional hand-coded approach to programming robots. Many past studies in imitation learning have human demonstrators perform tasks in the real world. However, this approach is generally expensive and requires high-qua...
متن کاملActive Imitation Learning via Reduction to I.I.D. Active Learning
In standard passive imitation learning, the goal is to learn a target policy by passively observing full execution trajectories of it. Unfortunately, generating such trajectories can require substantial expert effort and be impractical in some cases. In this paper, we consider active imitation learning with the goal of reducing this effort by querying the expert about the desired action at indi...
متن کاملActive lmitation learning: formal and practical reductions to I.I.D. learning
In standard passive imitation learning, the goal is to learn a policy that performs as well as a target policy by passively observing full execution trajectories of it. Unfortunately, generating such trajectories can require substantial expert effort and be impractical in some cases. In this paper, we consider active imitation learning with the goal of reducing this effort by querying the exper...
متن کاملModel-based Adversarial Imitation Learning
Generative adversarial learning is a popular new approach to training generative models which has been proven successful for other related problems as well. The general idea is to maintain an oracle D that discriminates between the expert’s data distribution and that of the generative model G. The generative model is trained to capture the expert’s distribution by maximizing the probability of ...
متن کاملDuring the First Three Months in Cross-Cultural Comparison The Emergence of Social Smiling: The Interplay of Maternal and Infant Imitation
The study addresses the emergence of the social smile in two different sociocultural contexts during the first 12 postnatal weeks. We examined different eliciting mechanisms like mutual gazing, maternal smile during mutual gazing, and reciprocal maternal and infant imitation of smiling. In co-constructivist theories of emotional development, all of them are considered social mechanisms that fos...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016